2025-06-08 17:56:28,991 [ 411127 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:42, check_args_and_update_paths) 2025-06-08 17:56:28,991 [ 411127 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:86, check_args_and_update_paths) 2025-06-08 17:56:28,991 [ 411127 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:97, check_args_and_update_paths) 2025-06-08 17:56:28,991 [ 411127 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:99, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_ikbgrs --privileged --dns-search='.' --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=2cffe1eae894 -e DOCKER_BASE_TAG=2993bc2bf171 -e DOCKER_KERBERIZED_HADOOP_TAG=ce74919e88f5 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=a2d3dc777d0c -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS=" -rfEps --run-id=1 --color=no --durations=0 test_postgresql_replica_database_engine_1/test.py::test_replicating_dml test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished test_postgresql_replica_database_engine_1/test.py::test_single_transaction test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots test_postgresql_replica_database_engine_1/test.py::test_virtual_columns -vvv" altinityinfra/integration-tests-runner:9d492c2eec24 '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: order-1.0.1, random-0.2, timeout-2.2.0, repeat-0.9.3, reportlog-0.4.0, xdist-3.5.0 timeout: 900.0s timeout method: signal timeout func_only: False collecting ... collected 6 items test_postgresql_replica_database_engine_1/test.py::test_replicating_dml FAILED [ 16%] test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished FAILED [ 33%] test_postgresql_replica_database_engine_1/test.py::test_single_transaction FAILED [ 50%] test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes FAILED [ 66%] test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots FAILED [ 83%] test_postgresql_replica_database_engine_1/test.py::test_virtual_columns FAILED [100%] =================================== FAILURES =================================== _____________________________ test_replicating_dml _____________________________ started_cluster = def test_replicating_dml(started_cluster): NUM_TABLES = 5 for i in range(NUM_TABLES): pg_manager.create_postgres_table(f"postgresql_replica_{i}") instance.query( "INSERT INTO postgres_database.postgresql_replica_{} SELECT number, {} from numbers(50)".format( i, i ) ) pg_manager.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port ) for i in range(NUM_TABLES): instance.query( f"INSERT INTO postgres_database.postgresql_replica_{i} SELECT 50 + number, {i} from numbers(1000)" ) > check_several_tables_are_synchronized(instance, NUM_TABLES) test_postgresql_replica_database_engine_1/test.py:100: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:419: in check_several_tables_are_synchronized check_tables_are_synchronized( helpers/postgres_utility.py:392: in check_tables_are_synchronized result = instance.query(result_query) helpers/cluster.py:3456: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 60, stderr: Received exception from server (version 24.3.18): E Code: 60. DB::Exception: Received from 172.16.1.3:9000. DB::Exception: Unknown table expression identifier 'test_database.postgresql_replica_0' in scope SELECT * FROM `test_database.postgresql_replica_0` ORDER BY key ASC. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000001d2e15a3 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000eff24b4 E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x0000000007d363fc E 3. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::(anonymous namespace)::QueryAnalyzer::resolveQuery(std::shared_ptr const&, DB::(anonymous namespace)::IdentifierResolveScope&) @ 0x00000000173a87a6 E 4. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::QueryAnalysisPass::run(std::shared_ptr&, std::shared_ptr) @ 0x00000000173a162f E 5. ./build_docker/./src/Analyzer/QueryTreePassManager.cpp:0: DB::QueryTreePassManager::run(std::shared_ptr) @ 0x000000001739ca9a E 6. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:0: DB::(anonymous namespace)::buildQueryTreeAndRunPasses(std::shared_ptr const&, DB::SelectQueryOptions const&, std::shared_ptr const&, std::shared_ptr const&) @ 0x0000000017862e10 E 7. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:160: DB::InterpreterSelectQueryAnalyzer::InterpreterSelectQueryAnalyzer(std::shared_ptr const&, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&) @ 0x000000001785f7e0 E 8. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:701: std::unique_ptr> std::__function::__policy_invoker> (DB::InterpreterFactory::Arguments const&)>::__call_impl> (DB::InterpreterFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::InterpreterFactory::Arguments const&) @ 0x0000000017864110 E 9. ./build_docker/./src/Interpreters/InterpreterFactory.cpp:0: DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x00000000177cfb6b E 10. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:302: DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x0000000017dbf13f E 11. ./build_docker/./src/Interpreters/executeQuery.cpp:1376: DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000017dbac38 E 12. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:612: DB::TCPHandler::runImpl() @ 0x00000000195edc92 E 13. ./build_docker/./src/Server/TCPHandler.cpp:2341: DB::TCPHandler::run() @ 0x000000001960c2e8 E 14. ./build_docker/./base/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x000000001d16e0a3 E 15. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x000000001d16e912 E 16. ./build_docker/./base/poco/Foundation/src/ThreadPool.cpp:202: Poco::PooledThread::run() @ 0x000000001d3705c7 E 17. ./build_docker/./base/poco/Foundation/src/Thread.cpp:46: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x000000001d36e890 E 18. ./base/poco/Foundation/include/Poco/SharedPtr.h:231: Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001d36cd4a E 19. __tsan_thread_start_func @ 0x000000000706b0cf E 20. ? @ 0x00007f2c70203ac3 E 21. ? @ 0x00007f2c70295850 E . (UNKNOWN_TABLE) E (query: select * from `test_database.postgresql_replica_0` order by key;) helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml PostgreSQL is available - running test ------------------------------ Captured log setup ------------------------------ 2025-06-08 17:56:31 [ 550 ] DEBUG : Command:['docker ps | wc -l'] (cluster.py:113, run_and_check) 2025-06-08 17:56:31 [ 550 ] DEBUG : Stdout:1 (cluster.py:121, run_and_check) 2025-06-08 17:56:31 [ 550 ] DEBUG : No running containers (conftest.py:92, cleanup_environment) 2025-06-08 17:56:31 [ 550 ] DEBUG : Pruning Docker networks (conftest.py:94, cleanup_environment) 2025-06-08 17:56:31 [ 550 ] DEBUG : Command:['docker network prune --force'] (cluster.py:113, run_and_check) 2025-06-08 17:56:31 [ 550 ] DEBUG : Command:["sysctl net.ipv4.ip_local_port_range='55000 65535'"] (cluster.py:113, run_and_check) 2025-06-08 17:56:31 [ 550 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:121, run_and_check) 2025-06-08 17:56:31 [ 550 ] INFO : Running tests in /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/test.py (cluster.py:2659, start) 2025-06-08 17:56:31 [ 550 ] DEBUG : Cluster start called. is_up=False (cluster.py:2666, start) 2025-06-08 17:56:31 [ 550 ] DEBUG : Docker networks for project roottestpostgresqlreplicadatabaseengine1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:780, print_all_docker_pieces) 2025-06-08 17:56:31 [ 550 ] DEBUG : Docker containers for project roottestpostgresqlreplicadatabaseengine1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:788, print_all_docker_pieces) 2025-06-08 17:56:31 [ 550 ] DEBUG : Docker volumes for project roottestpostgresqlreplicadatabaseengine1 are DRIVER VOLUME NAME (cluster.py:796, print_all_docker_pieces) 2025-06-08 17:56:31 [ 550 ] DEBUG : Cleanup called (cluster.py:801, cleanup) 2025-06-08 17:56:31 [ 550 ] DEBUG : Docker networks for project roottestpostgresqlreplicadatabaseengine1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:780, print_all_docker_pieces) 2025-06-08 17:56:32 [ 550 ] DEBUG : Docker containers for project roottestpostgresqlreplicadatabaseengine1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:788, print_all_docker_pieces) 2025-06-08 17:56:32 [ 550 ] DEBUG : Docker volumes for project roottestpostgresqlreplicadatabaseengine1 are DRIVER VOLUME NAME (cluster.py:796, print_all_docker_pieces) 2025-06-08 17:56:32 [ 550 ] DEBUG : Command:docker container list --all --filter name='^/roottestpostgresqlreplicadatabaseengine1_.*_1$' --format '{{.ID}}:{{.Names}}' (cluster.py:113, run_and_check) 2025-06-08 17:56:32 [ 550 ] DEBUG : Unstopped containers: {} (cluster.py:815, cleanup) 2025-06-08 17:56:32 [ 550 ] DEBUG : No running containers for project: roottestpostgresqlreplicadatabaseengine1 (cluster.py:829, cleanup) 2025-06-08 17:56:32 [ 550 ] DEBUG : Trying to prune unused networks... (cluster.py:835, cleanup) 2025-06-08 17:56:32 [ 550 ] DEBUG : Trying to prune unused images... (cluster.py:851, cleanup) 2025-06-08 17:56:32 [ 550 ] DEBUG : Command:['docker', 'image', 'prune', '-f'] (cluster.py:113, run_and_check) 2025-06-08 17:56:32 [ 550 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:121, run_and_check) 2025-06-08 17:56:32 [ 550 ] DEBUG : Images pruned (cluster.py:854, cleanup) 2025-06-08 17:56:32 [ 550 ] DEBUG : Trying to prune unused volumes... (cluster.py:860, cleanup) 2025-06-08 17:56:32 [ 550 ] DEBUG : Command:['docker volume ls | wc -l'] (cluster.py:113, run_and_check) 2025-06-08 17:56:32 [ 550 ] DEBUG : Stdout:1 (cluster.py:121, run_and_check) 2025-06-08 17:56:32 [ 550 ] DEBUG : Setup directory for instance: instance (cluster.py:2679, start) 2025-06-08 17:56:32 [ 550 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4383, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Create directory for common tests configuration (cluster.py:4388, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Copy common configuration from helpers (cluster.py:4408, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Generate and write macros file (cluster.py:4441, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/configs/log_conf.xml'] to /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/configs/config.d (cluster.py:4471, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/database (cluster.py:4488, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/logs (cluster.py:4499, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Entrypoint cmd: bash -c "trap 'pkill tail' INT TERM; clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon; coproc tail -f /dev/null; wait $$!" (cluster.py:4582, create_dir) 2025-06-08 17:56:32 [ 550 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'POSTGRES_PORT': '5432', 'POSTGRES_DIR': '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/postgres/postgres1', 'POSTGRES_LOGS_FS': 'bind'} stored in /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env (cluster.py:86, _create_env_file) 2025-06-08 17:56:32 [ 550 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-06-08 17:56:32 [ 550 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-06-08 17:56:32 [ 550 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-06-08 17:56:32 [ 550 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-06-08 17:56:32 [ 550 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:546, _make_request) 2025-06-08 17:56:32 [ 550 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine1', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', 'pull'] (cluster.py:113, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling instance ... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling postgres1 ... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling instance ... pulling from altinityinfra/integr... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling instance ... digest: sha256:8a2c68e2d63d82c826... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling instance ... status: image is up to date for a... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling instance ... done (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling postgres1 ... pulling from library/postgres (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling postgres1 ... digest: sha256:6efd0df010dc3cb40d... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling postgres1 ... status: image is up to date for p... (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Stderr:Pulling postgres1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:56:42 [ 550 ] DEBUG : Setup Postgres (cluster.py:2791, start) 2025-06-08 17:56:42 [ 550 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine1', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', '--verbose', 'up', '-d'] (cluster.py:113, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.config.config.find: Using configuration files: /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.docker_client.get_client: docker-compose version 1.29.2, build unknown (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:docker-py version: (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:CPython version: 3.10.12 (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:OpenSSL version: OpenSSL 3.0.2 15 Mar 2022 (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.docker_client.get_client: Docker base_url: http+docker://localhost (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.docker_client.get_client: Docker version: Platform={'Name': 'Docker Engine - Community'}, Components=[{'Name': 'Engine', 'Version': '23.0.6', 'Details': {'ApiVersion': '1.42', 'Arch': 'amd64', 'BuildTime': '2023-05-05T21:18:13.000000000+00:00', 'Experimental': 'false', 'GitCommit': '9dbdbd4', 'GoVersion': 'go1.19.9', 'KernelVersion': '5.15.0-130-generic', 'MinAPIVersion': '1.12', 'Os': 'linux'}}, {'Name': 'containerd', 'Version': '1.7.18', 'Details': {'GitCommit': 'ae71819c4f5e67bb4d5ae76a6b735f29cc25774e'}}, {'Name': 'runc', 'Version': '1.7.18', 'Details': {'GitCommit': 'v1.1.13-0-g58aa920'}}, {'Name': 'docker-init', 'Version': '0.19.0', 'Details': {'GitCommit': 'de40ad0'}}], Version=23.0.6, ApiVersion=1.42, MinAPIVersion=1.12, GitCommit=9dbdbd4, GoVersion=go1.19.9, Os=linux, Arch=amd64, KernelVersion=5.15.0-130-generic, BuildTime=2023-05-05T21:18:13.000000000+00:00 (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_network <- ('roottestpostgresqlreplicadatabaseengine1_default') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker info <- () (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker info -> {'Architecture': 'x86_64', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'BridgeNfIp6tables': True, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'BridgeNfIptables': True, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'CPUSet': True, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'CPUShares': True, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'CgroupDriver': 'cgroupfs', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'CgroupVersion': '2', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'ContainerdCommit': {'Expected': 'ae71819c4f5e67bb4d5ae76a6b735f29cc25774e', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'ID': 'ae71819c4f5e67bb4d5ae76a6b735f29cc25774e'}, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Containers': 0, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_network <- ('roottestpostgresqlreplicadatabaseengine1_default') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.network.ensure: Creating network "roottestpostgresqlreplicadatabaseengine1_default" with the default driver (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker create_network <- (name='roottestpostgresqlreplicadatabaseengine1_default', driver=None, options=None, ipam=None, internal=False, enable_ipv6=False, labels={'com.docker.compose.project': 'roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.network': 'default', 'com.docker.compose.version': '1.29.2'}, attachable=True, check_duplicate=True) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker create_network -> {'Id': '132dfdacc4017067c4f403852aa07d55616ee50e234a2ee23cba059ed949db95', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Warning': ''} (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=False, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=False, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service=postgres1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service=postgres1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_image <- ('postgres') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_image -> {'Architecture': 'amd64', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Author': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Comment': 'buildkit.dockerfile.v0', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Config': {'AttachStderr': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'AttachStdin': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'AttachStdout': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Cmd': ['postgres'], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Domainname': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Entrypoint': ['docker-entrypoint.sh'], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Env': ['PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/lib/postgresql/17/bin', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service=postgres1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service=postgres1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: {} (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Starting producer thread for (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service=postgres1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers <- (all=True, filters={'label': ['com.docker.compose.project=roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service=postgres1', 'com.docker.compose.oneoff=False']}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker containers -> (list with 0 items) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:Creating roottestpostgresqlreplicadatabaseengine1_postgres1_1 ... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: {ServiceName(project='roottestpostgresqlreplicadatabaseengine1', service='postgres1', number=1)} (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Starting producer thread for ServiceName(project='roottestpostgresqlreplicadatabaseengine1', service='postgres1', number=1) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_image <- ('postgres') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_image -> {'Architecture': 'amd64', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Author': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Comment': 'buildkit.dockerfile.v0', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Config': {'AttachStderr': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'AttachStdin': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'AttachStdout': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Cmd': ['postgres'], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Domainname': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Entrypoint': ['docker-entrypoint.sh'], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Env': ['PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/lib/postgresql/17/bin', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_image <- ('postgres') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_image -> {'Architecture': 'amd64', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Author': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Comment': 'buildkit.dockerfile.v0', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Config': {'AttachStderr': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'AttachStdin': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'AttachStdout': False, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Cmd': ['postgres'], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Domainname': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Entrypoint': ['docker-entrypoint.sh'], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Env': ['PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/usr/lib/postgresql/17/bin', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.service.build_container_labels: Added config hash: 9feb8ae4c4f839ef6d94321052b7e6360a3c957f2125293b3a72d12ae3978768 (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker create_host_config <- (links=[], port_bindings={}, binds=[], volumes_from=[], privileged=False, network_mode='roottestpostgresqlreplicadatabaseengine1_default', devices=None, device_requests=None, dns=None, dns_opt=None, dns_search=None, restart_policy={'Name': 'always', 'MaximumRetryCount': 0}, runtime=None, cap_add=None, cap_drop=None, mem_limit=None, mem_reservation=None, memswap_limit=None, ulimits=None, log_config={'Type': '', 'Config': {}}, extra_hosts=None, read_only=None, pid_mode=None, security_opt=None, ipc_mode=None, cgroup_parent=None, cpu_quota=None, shm_size=None, sysctls=None, pids_limit=None, tmpfs=None, oom_kill_disable=None, oom_score_adj=None, mem_swappiness=None, group_add=None, userns_mode=None, init=None, init_path=None, isolation=None, cpu_count=None, cpu_percent=None, nano_cpus=None, volume_driver=None, cpuset_cpus=None, cpu_shares=None, storage_opt=None, blkio_weight=None, blkio_weight_device=None, device_read_bps=None, device_read_iops=None, device_write_bps=None, device_write_iops=None, mounts=[{'Target': '/postgres/', 'Source': '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/postgres/postgres1', 'Type': 'bind', 'ReadOnly': None}], device_cgroup_rules=None, cpu_period=None, cpu_rt_period=None, cpu_rt_runtime=None) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker create_host_config -> {'Binds': [], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Links': [], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'LogConfig': {'Config': {}, 'Type': ''}, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Mounts': [{'ReadOnly': None, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Source': '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/postgres/postgres1', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Target': '/postgres/', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Type': 'bind'}], (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'NetworkMode': 'roottestpostgresqlreplicadatabaseengine1_default', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'PortBindings': {}, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'RestartPolicy': {'MaximumRetryCount': 0, 'Name': 'always'}, (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker create_container <- (command=['postgres', '-c', 'wal_level=logical', '-c', 'max_replication_slots=4', '-c', 'logging_collector=on', '-c', 'log_directory=/postgres/logs', '-c', 'log_filename=postgresql.log', '-c', 'log_statement=all', '-c', 'max_connections=200'], environment=['POSTGRES_HOST_AUTH_METHOD=trust', 'POSTGRES_PASSWORD=mysecretpassword', 'PGDATA=/postgres/data'], healthcheck={'test': ['CMD-SHELL', 'pg_isready -U postgres'], 'interval': 10000000000, 'timeout': 5000000000, 'retries': 5}, image='postgres', volumes={}, name='roottestpostgresqlreplicadatabaseengine1_postgres1_1', detach=True, ports=['5432'], labels={'com.docker.compose.project': 'roottestpostgresqlreplicadatabaseengine1', 'com.docker.compose.service': 'postgres1', 'com.docker.compose.oneoff': 'False', 'com.docker.compose.project.working_dir': '/ClickHouse/tests/integration/compose', 'com.docker.compose.project.config_files': '/ClickHouse/tests/integration/compose/docker_compose_postgres.yml', 'com.docker.compose.project.environment_file': '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env', 'com.docker.compose.container-number': '1', 'com.docker.compose.version': '1.29.2', 'com.docker.compose.config-hash': '9feb8ae4c4f839ef6d94321052b7e6360a3c957f2125293b3a72d12ae3978768'}, host_config={'NetworkMode': 'roottestpostgresqlreplicadatabaseengine1_default', 'RestartPolicy': {'Name': 'always', 'MaximumRetryCount': 0}, 'VolumesFrom': [], 'Binds': [], 'PortBindings': {}, 'Links': [], 'LogConfig': {'Type': '', 'Config': {}}, 'Mounts': [{'Target': '/postgres/', 'Source': '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/postgres/postgres1', 'Type': 'bind', 'ReadOnly': None}]}, networking_config={'EndpointsConfig': {'roottestpostgresqlreplicadatabaseengine1_default': {'Aliases': ['postgres1', 'postgre-sql.local'], 'IPAMConfig': {}}}}) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker create_container -> {'Id': '2cc4eade8082f3f5694c66648547993fe385dd4a29be025d1a0df00a9d24d2a8', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Warnings': []} (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_container <- ('2cc4eade8082f3f5694c66648547993fe385dd4a29be025d1a0df00a9d24d2a8') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker inspect_container -> {'AppArmorProfile': '', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'Args': ['postgres', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: '-c', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'wal_level=logical', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: '-c', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'max_replication_slots=4', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: '-c', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'logging_collector=on', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: '-c', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr: 'log_directory=/postgres/logs', (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:... (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker disconnect_container_from_network <- ('2cc4eade8082f3f5694c66648547993fe385dd4a29be025d1a0df00a9d24d2a8', 'roottestpostgresqlreplicadatabaseengine1_default') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker disconnect_container_from_network -> None (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker connect_container_to_network <- ('2cc4eade8082f3f5694c66648547993fe385dd4a29be025d1a0df00a9d24d2a8', 'roottestpostgresqlreplicadatabaseengine1_default', aliases=['postgres1', '2cc4eade8082', 'postgre-sql.local'], ipv4_address=None, ipv6_address=None, links=[], link_local_ips=None) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker connect_container_to_network -> None (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker start <- ('2cc4eade8082f3f5694c66648547993fe385dd4a29be025d1a0df00a9d24d2a8') (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.cli.verbose_proxy.proxy_callable: docker start -> None (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.parallel_execute_iter: Finished processing: ServiceName(project='roottestpostgresqlreplicadatabaseengine1', service='postgres1', number=1) (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:Creating roottestpostgresqlreplicadatabaseengine1_postgres1_1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.parallel_execute_iter: Finished processing: (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : Stderr:compose.parallel.feed_queue: Pending: set() (cluster.py:123, run_and_check) 2025-06-08 17:56:43 [ 550 ] DEBUG : get_instance_ip instance_name=postgres1 (cluster.py:2008, get_instance_ip) 2025-06-08 17:56:43 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestpostgresqlreplicadatabaseengine1_postgres1_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:43 [ 550 ] DEBUG : Can't connect to Postgres connection to server at "172.16.1.2", port 5432 failed: Connection refused Is the server running on that host and accepting TCP/IP connections? (cluster.py:2251, wait_postgres_to_start) 2025-06-08 17:56:44 [ 550 ] DEBUG : Can't connect to Postgres connection to server at "172.16.1.2", port 5432 failed: Connection refused Is the server running on that host and accepting TCP/IP connections? (cluster.py:2251, wait_postgres_to_start) 2025-06-08 17:56:44 [ 550 ] DEBUG : Postgres Started (cluster.py:2248, wait_postgres_to_start) 2025-06-08 17:56:44 [ 550 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker-compose --env-file /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env --project-name roottestpostgresqlreplicadatabaseengine1 --file /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml up -d --no-recreate') (cluster.py:3002, start) 2025-06-08 17:56:44 [ 550 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine1', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', 'up', '-d', '--no-recreate'] (cluster.py:113, run_and_check) 2025-06-08 17:56:45 [ 550 ] DEBUG : Stderr:Creating roottestpostgresqlreplicadatabaseengine1_instance_1 ... (cluster.py:123, run_and_check) 2025-06-08 17:56:45 [ 550 ] DEBUG : Stderr:Creating roottestpostgresqlreplicadatabaseengine1_instance_1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:56:45 [ 550 ] DEBUG : ClickHouse instance created (cluster.py:3010, start) 2025-06-08 17:56:45 [ 550 ] DEBUG : get_instance_ip instance_name=instance (cluster.py:2008, get_instance_ip) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestpostgresqlreplicadatabaseengine1_instance_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : Waiting for ClickHouse start in instance, ip: 172.16.1.3... (cluster.py:3017, start) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/roottestpostgresqlreplicadatabaseengine1_instance_1/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:45 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:46 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:47 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:47 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:47 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:47 [ 550 ] DEBUG : http://localhost:None "GET /v1.42/containers/223b92849121654602cb2530af1005aea2232a14ba947478c3700ea6f9210425/json HTTP/1.1" 200 None (connectionpool.py:546, _make_request) 2025-06-08 17:56:47 [ 550 ] DEBUG : ClickHouse instance started (cluster.py:3021, start) 2025-06-08 17:56:47 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:56:47 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_0" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_1" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_2" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_3" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_4" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Checking table postgresql_replica_0 exists in test_database Checking table is synchronized: test_database.postgresql_replica_0 ------------------------------ Captured log call ------------------------------- 2025-06-08 17:56:47 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_0 SELECT number, 0 from numbers(50) on instance (cluster.py:3455, query) 2025-06-08 17:56:48 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_1 SELECT number, 1 from numbers(50) on instance (cluster.py:3455, query) 2025-06-08 17:56:48 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_2 SELECT number, 2 from numbers(50) on instance (cluster.py:3455, query) 2025-06-08 17:56:48 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_3 SELECT number, 3 from numbers(50) on instance (cluster.py:3455, query) 2025-06-08 17:56:48 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_4 SELECT number, 4 from numbers(50) on instance (cluster.py:3455, query) 2025-06-08 17:56:49 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:56:49 [ 550 ] DEBUG : Executing query CREATE DATABASE `test_database` ENGINE = MaterializedPostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) 2025-06-08 17:56:49 [ 550 ] DEBUG : Executing query SHOW DATABASES on instance (cluster.py:3455, query) 2025-06-08 17:56:49 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_0 SELECT 50 + number, 0 from numbers(1000) on instance (cluster.py:3455, query) 2025-06-08 17:56:49 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_1 SELECT 50 + number, 1 from numbers(1000) on instance (cluster.py:3455, query) 2025-06-08 17:56:50 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_2 SELECT 50 + number, 2 from numbers(1000) on instance (cluster.py:3455, query) 2025-06-08 17:56:50 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_3 SELECT 50 + number, 3 from numbers(1000) on instance (cluster.py:3455, query) 2025-06-08 17:56:50 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_4 SELECT 50 + number, 4 from numbers(1000) on instance (cluster.py:3455, query) 2025-06-08 17:56:50 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:56:50 [ 550 ] DEBUG : Executing query select * from `postgres_database`.`postgresql_replica_0` order by key; on instance (cluster.py:3455, query) 2025-06-08 17:56:51 [ 550 ] DEBUG : Executing query select * from `test_database.postgresql_replica_0` order by key; on instance (cluster.py:3455, query) ---------------------------- Captured log teardown ----------------------------- 2025-06-08 17:56:51 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` SYNC on instance (cluster.py:3455, query) 2025-06-08 17:56:51 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:56:51 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:56:52 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) __________ test_restart_server_while_replication_startup_not_finished __________ started_cluster = def test_restart_server_while_replication_startup_not_finished(started_cluster): NUM_TABLES = 5 pg_manager.create_and_fill_postgres_tables(NUM_TABLES, 100000) pg_manager.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port ) time.sleep(1) instance.restart_clickhouse() > check_several_tables_are_synchronized(instance, NUM_TABLES) test_postgresql_replica_database_engine_1/test.py:774: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:419: in check_several_tables_are_synchronized check_tables_are_synchronized( helpers/postgres_utility.py:392: in check_tables_are_synchronized result = instance.query(result_query) helpers/cluster.py:3456: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 60, stderr: Received exception from server (version 24.3.18): E Code: 60. DB::Exception: Received from 172.16.1.3:9000. DB::Exception: Unknown table expression identifier 'test_database.postgresql_replica_0' in scope SELECT * FROM `test_database.postgresql_replica_0` ORDER BY key ASC. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000001d2e15a3 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000eff24b4 E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x0000000007d363fc E 3. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::(anonymous namespace)::QueryAnalyzer::resolveQuery(std::shared_ptr const&, DB::(anonymous namespace)::IdentifierResolveScope&) @ 0x00000000173a87a6 E 4. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::QueryAnalysisPass::run(std::shared_ptr&, std::shared_ptr) @ 0x00000000173a162f E 5. ./build_docker/./src/Analyzer/QueryTreePassManager.cpp:0: DB::QueryTreePassManager::run(std::shared_ptr) @ 0x000000001739ca9a E 6. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:0: DB::(anonymous namespace)::buildQueryTreeAndRunPasses(std::shared_ptr const&, DB::SelectQueryOptions const&, std::shared_ptr const&, std::shared_ptr const&) @ 0x0000000017862e10 E 7. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:160: DB::InterpreterSelectQueryAnalyzer::InterpreterSelectQueryAnalyzer(std::shared_ptr const&, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&) @ 0x000000001785f7e0 E 8. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:701: std::unique_ptr> std::__function::__policy_invoker> (DB::InterpreterFactory::Arguments const&)>::__call_impl> (DB::InterpreterFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::InterpreterFactory::Arguments const&) @ 0x0000000017864110 E 9. ./build_docker/./src/Interpreters/InterpreterFactory.cpp:0: DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x00000000177cfb6b E 10. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:302: DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x0000000017dbf13f E 11. ./build_docker/./src/Interpreters/executeQuery.cpp:1376: DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000017dbac38 E 12. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:612: DB::TCPHandler::runImpl() @ 0x00000000195edc92 E 13. ./build_docker/./src/Server/TCPHandler.cpp:2341: DB::TCPHandler::run() @ 0x000000001960c2e8 E 14. ./build_docker/./base/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x000000001d16e0a3 E 15. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x000000001d16e912 E 16. ./build_docker/./base/poco/Foundation/src/ThreadPool.cpp:202: Poco::PooledThread::run() @ 0x000000001d3705c7 E 17. ./build_docker/./base/poco/Foundation/src/Thread.cpp:46: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x000000001d36e890 E 18. ./base/poco/Foundation/include/Poco/SharedPtr.h:231: Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001d36cd4a E 19. __tsan_thread_start_func @ 0x000000000706b0cf E 20. ? @ 0x00007fbf65c25ac3 E 21. ? @ 0x00007fbf65cb7850 E . (UNKNOWN_TABLE) E (query: select * from `test_database.postgresql_replica_0` order by key;) helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- PostgreSQL is available - running test ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_0" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_1" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_2" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_3" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_4" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Checking table postgresql_replica_0 exists in test_database Checking table is synchronized: test_database.postgresql_replica_0 ------------------------------ Captured log call ------------------------------- 2025-06-08 17:56:52 [ 550 ] DEBUG : Executing query INSERT INTO `postgres_database`.`postgresql_replica_0` SELECT number, number from numbers(100000) on instance (cluster.py:3455, query) 2025-06-08 17:56:53 [ 550 ] DEBUG : Executing query INSERT INTO `postgres_database`.`postgresql_replica_1` SELECT number, number from numbers(100000) on instance (cluster.py:3455, query) 2025-06-08 17:56:54 [ 550 ] DEBUG : Executing query INSERT INTO `postgres_database`.`postgresql_replica_2` SELECT number, number from numbers(100000) on instance (cluster.py:3455, query) 2025-06-08 17:56:55 [ 550 ] DEBUG : Executing query INSERT INTO `postgres_database`.`postgresql_replica_3` SELECT number, number from numbers(100000) on instance (cluster.py:3455, query) 2025-06-08 17:56:56 [ 550 ] DEBUG : Executing query INSERT INTO `postgres_database`.`postgresql_replica_4` SELECT number, number from numbers(100000) on instance (cluster.py:3455, query) 2025-06-08 17:56:57 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:56:58 [ 550 ] DEBUG : Executing query CREATE DATABASE `test_database` ENGINE = MaterializedPostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) 2025-06-08 17:56:58 [ 550 ] DEBUG : Executing query SHOW DATABASES on instance (cluster.py:3455, query) 2025-06-08 17:56:59 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:True cmd: ['bash', '-c', 'ps -C clickhouse'] (cluster.py:2046, exec_in_container) 2025-06-08 17:56:59 [ 550 ] DEBUG : Command:['docker', 'exec', '-u', 'root', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', 'ps -C clickhouse'] (cluster.py:113, run_and_check) 2025-06-08 17:56:59 [ 550 ] DEBUG : Stdout: PID TTY TIME CMD (cluster.py:121, run_and_check) 2025-06-08 17:56:59 [ 550 ] DEBUG : Stdout: 8 ? 00:00:10 clickhouse (cluster.py:121, run_and_check) 2025-06-08 17:56:59 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', 'pkill clickhouse'] (cluster.py:2046, exec_in_container) 2025-06-08 17:56:59 [ 550 ] DEBUG : Command:['docker', 'exec', '-u', 'root', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', 'pkill clickhouse'] (cluster.py:113, run_and_check) 2025-06-08 17:56:59 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2046, exec_in_container) 2025-06-08 17:56:59 [ 550 ] DEBUG : Command:['docker', 'exec', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:113, run_and_check) 2025-06-08 17:56:59 [ 550 ] DEBUG : Stdout:8 (cluster.py:121, run_and_check) 2025-06-08 17:57:00 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2046, exec_in_container) 2025-06-08 17:57:00 [ 550 ] DEBUG : Command:['docker', 'exec', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:113, run_and_check) 2025-06-08 17:57:00 [ 550 ] DEBUG : Stdout:8 (cluster.py:121, run_and_check) 2025-06-08 17:57:01 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2046, exec_in_container) 2025-06-08 17:57:01 [ 550 ] DEBUG : Command:['docker', 'exec', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:113, run_and_check) 2025-06-08 17:57:01 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2046, exec_in_container) 2025-06-08 17:57:01 [ 550 ] DEBUG : Command:['docker', 'exec', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:113, run_and_check) 2025-06-08 17:57:01 [ 550 ] DEBUG : No clickhouse process running. Start new one. (cluster.py:3817, start_clickhouse) 2025-06-08 17:57:01 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', 'clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon'] (cluster.py:2046, exec_in_container) 2025-06-08 17:57:01 [ 550 ] DEBUG : Command:['docker', 'exec', '-u', '0', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', 'clickhouse server --config-file=/etc/clickhouse-server/config.xml --log-file=/var/log/clickhouse-server/clickhouse-server.log --errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log --daemon'] (cluster.py:113, run_and_check) 2025-06-08 17:57:03 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2046, exec_in_container) 2025-06-08 17:57:03 [ 550 ] DEBUG : Command:['docker', 'exec', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:113, run_and_check) 2025-06-08 17:57:03 [ 550 ] DEBUG : Stdout:753 (cluster.py:121, run_and_check) 2025-06-08 17:57:03 [ 550 ] DEBUG : Clickhouse process running. (cluster.py:3828, start_clickhouse) 2025-06-08 17:57:03 [ 550 ] DEBUG : run container_id:roottestpostgresqlreplicadatabaseengine1_instance_1 detach:False nothrow:False cmd: ['bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:2046, exec_in_container) 2025-06-08 17:57:03 [ 550 ] DEBUG : Command:['docker', 'exec', 'roottestpostgresqlreplicadatabaseengine1_instance_1', 'bash', '-c', "ps ax | grep 'clickhouse' | grep -v 'grep' | grep -v 'coproc' | grep -v 'bash -c' | awk '{print $1}'"] (cluster.py:113, run_and_check) 2025-06-08 17:57:03 [ 550 ] DEBUG : Stdout:753 (cluster.py:121, run_and_check) 2025-06-08 17:57:03 [ 550 ] DEBUG : Executing query select 20 on instance (cluster.py:3455, query) 2025-06-08 17:57:03 [ 550 ] DEBUG : Executing query select 20 on instance (cluster.py:3455, query) 2025-06-08 17:57:04 [ 550 ] DEBUG : Executing query select 20 on instance (cluster.py:3455, query) 2025-06-08 17:57:04 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:04 [ 550 ] DEBUG : Executing query select * from `postgres_database`.`postgresql_replica_0` order by key; on instance (cluster.py:3455, query) 2025-06-08 17:57:05 [ 550 ] DEBUG : Executing query select * from `test_database.postgresql_replica_0` order by key; on instance (cluster.py:3455, query) ---------------------------- Captured log teardown ----------------------------- 2025-06-08 17:57:05 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` SYNC on instance (cluster.py:3455, query) 2025-06-08 17:57:05 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:06 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:06 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) ___________________________ test_single_transaction ____________________________ started_cluster = def test_single_transaction(started_cluster): conn = get_postgres_conn( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, database=True, auto_commit=False, ) cursor = conn.cursor() table_name = "postgresql_replica_0" create_postgres_table(cursor, table_name) conn.commit() pg_manager.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port ) assert_nested_table_is_created(instance, table_name) for query in queries: print("query {}".format(query)) cursor.execute(query.format(0)) time.sleep(5) result = instance.query(f"select count() from test_database.{table_name}") # no commit yet assert int(result) == 0 conn.commit() > check_tables_are_synchronized(instance, table_name) test_postgresql_replica_database_engine_1/test.py:531: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:392: in check_tables_are_synchronized result = instance.query(result_query) helpers/cluster.py:3456: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 60, stderr: Received exception from server (version 24.3.18): E Code: 60. DB::Exception: Received from 172.16.1.3:9000. DB::Exception: Unknown table expression identifier 'test_database.postgresql_replica_0' in scope SELECT * FROM `test_database.postgresql_replica_0` ORDER BY key ASC. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000001d2e15a3 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000eff24b4 E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x0000000007d363fc E 3. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::(anonymous namespace)::QueryAnalyzer::resolveQuery(std::shared_ptr const&, DB::(anonymous namespace)::IdentifierResolveScope&) @ 0x00000000173a87a6 E 4. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::QueryAnalysisPass::run(std::shared_ptr&, std::shared_ptr) @ 0x00000000173a162f E 5. ./build_docker/./src/Analyzer/QueryTreePassManager.cpp:0: DB::QueryTreePassManager::run(std::shared_ptr) @ 0x000000001739ca9a E 6. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:0: DB::(anonymous namespace)::buildQueryTreeAndRunPasses(std::shared_ptr const&, DB::SelectQueryOptions const&, std::shared_ptr const&, std::shared_ptr const&) @ 0x0000000017862e10 E 7. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:160: DB::InterpreterSelectQueryAnalyzer::InterpreterSelectQueryAnalyzer(std::shared_ptr const&, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&) @ 0x000000001785f7e0 E 8. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:701: std::unique_ptr> std::__function::__policy_invoker> (DB::InterpreterFactory::Arguments const&)>::__call_impl> (DB::InterpreterFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::InterpreterFactory::Arguments const&) @ 0x0000000017864110 E 9. ./build_docker/./src/Interpreters/InterpreterFactory.cpp:0: DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x00000000177cfb6b E 10. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:302: DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x0000000017dbf13f E 11. ./build_docker/./src/Interpreters/executeQuery.cpp:1376: DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000017dbac38 E 12. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:612: DB::TCPHandler::runImpl() @ 0x00000000195edc92 E 13. ./build_docker/./src/Server/TCPHandler.cpp:2341: DB::TCPHandler::run() @ 0x000000001960c2e8 E 14. ./build_docker/./base/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x000000001d16e0a3 E 15. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x000000001d16e912 E 16. ./build_docker/./base/poco/Foundation/src/ThreadPool.cpp:202: Poco::PooledThread::run() @ 0x000000001d3705c7 E 17. ./build_docker/./base/poco/Foundation/src/Thread.cpp:46: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x000000001d36e890 E 18. ./base/poco/Foundation/include/Poco/SharedPtr.h:231: Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001d36cd4a E 19. __tsan_thread_start_func @ 0x000000000706b0cf E 20. ? @ 0x00007fbf65c25ac3 E 21. ? @ 0x00007fbf65cb7850 E . (UNKNOWN_TABLE) E (query: select * from `test_database.postgresql_replica_0` order by key;) helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- PostgreSQL is available - running test ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_0" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Checking table postgresql_replica_0 exists in test_database query INSERT INTO postgresql_replica_{} select i, i from generate_series(0, 10000) as t(i); query DELETE FROM postgresql_replica_{} WHERE (value*value) % 3 = 0; query UPDATE postgresql_replica_{} SET value = value - 125 WHERE key % 2 = 0; query UPDATE postgresql_replica_{} SET key=key+20000 WHERE key%2=0 query INSERT INTO postgresql_replica_{} select i, i from generate_series(40000, 50000) as t(i); query DELETE FROM postgresql_replica_{} WHERE key % 10 = 0; query UPDATE postgresql_replica_{} SET value = value + 101 WHERE key % 2 = 1; query UPDATE postgresql_replica_{} SET key=key+80000 WHERE key%2=1 query DELETE FROM postgresql_replica_{} WHERE value % 2 = 0; query UPDATE postgresql_replica_{} SET value = value + 2000 WHERE key % 5 = 0; query INSERT INTO postgresql_replica_{} select i, i from generate_series(200000, 250000) as t(i); query DELETE FROM postgresql_replica_{} WHERE value % 3 = 0; query UPDATE postgresql_replica_{} SET value = value * 2 WHERE key % 3 = 0; query UPDATE postgresql_replica_{} SET key=key+500000 WHERE key%2=1 query INSERT INTO postgresql_replica_{} select i, i from generate_series(1000000, 1050000) as t(i); query DELETE FROM postgresql_replica_{} WHERE value % 9 = 2; query UPDATE postgresql_replica_{} SET key=key+10000000 query UPDATE postgresql_replica_{} SET value = value + 2 WHERE key % 3 = 1; query DELETE FROM postgresql_replica_{} WHERE value%5 = 0; Checking table postgresql_replica_0 exists in test_database Checking table is synchronized: test_database.postgresql_replica_0 ------------------------------ Captured log call ------------------------------- 2025-06-08 17:57:06 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:06 [ 550 ] DEBUG : Executing query CREATE DATABASE `test_database` ENGINE = MaterializedPostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) 2025-06-08 17:57:06 [ 550 ] DEBUG : Executing query SHOW DATABASES on instance (cluster.py:3455, query) 2025-06-08 17:57:07 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:12 [ 550 ] DEBUG : Executing query select count() from test_database.postgresql_replica_0 on instance (cluster.py:3455, query) 2025-06-08 17:57:13 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:13 [ 550 ] DEBUG : Executing query select * from `postgres_database`.`postgresql_replica_0` order by key; on instance (cluster.py:3455, query) 2025-06-08 17:57:13 [ 550 ] DEBUG : Executing query select * from `test_database.postgresql_replica_0` order by key; on instance (cluster.py:3455, query) ---------------------------- Captured log teardown ----------------------------- 2025-06-08 17:57:14 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` SYNC on instance (cluster.py:3455, query) 2025-06-08 17:57:14 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:14 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:14 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) __________________________ test_table_schema_changes ___________________________ started_cluster = def test_table_schema_changes(started_cluster): NUM_TABLES = 5 for i in range(NUM_TABLES): pg_manager.create_postgres_table( f"postgresql_replica_{i}", template=postgres_table_template_2 ) instance.query( f"INSERT INTO postgres_database.postgresql_replica_{i} SELECT number, {i}, {i}, {i} from numbers(25)" ) pg_manager.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, ) for i in range(NUM_TABLES): instance.query( f"INSERT INTO postgres_database.postgresql_replica_{i} SELECT 25 + number, {i}, {i}, {i} from numbers(25)" ) > check_several_tables_are_synchronized(instance, NUM_TABLES) test_postgresql_replica_database_engine_1/test.py:371: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:419: in check_several_tables_are_synchronized check_tables_are_synchronized( helpers/postgres_utility.py:392: in check_tables_are_synchronized result = instance.query(result_query) helpers/cluster.py:3456: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 60, stderr: Received exception from server (version 24.3.18): E Code: 60. DB::Exception: Received from 172.16.1.3:9000. DB::Exception: Unknown table expression identifier 'test_database.postgresql_replica_0' in scope SELECT * FROM `test_database.postgresql_replica_0` ORDER BY key ASC. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000001d2e15a3 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000eff24b4 E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x0000000007d363fc E 3. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::(anonymous namespace)::QueryAnalyzer::resolveQuery(std::shared_ptr const&, DB::(anonymous namespace)::IdentifierResolveScope&) @ 0x00000000173a87a6 E 4. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::QueryAnalysisPass::run(std::shared_ptr&, std::shared_ptr) @ 0x00000000173a162f E 5. ./build_docker/./src/Analyzer/QueryTreePassManager.cpp:0: DB::QueryTreePassManager::run(std::shared_ptr) @ 0x000000001739ca9a E 6. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:0: DB::(anonymous namespace)::buildQueryTreeAndRunPasses(std::shared_ptr const&, DB::SelectQueryOptions const&, std::shared_ptr const&, std::shared_ptr const&) @ 0x0000000017862e10 E 7. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:160: DB::InterpreterSelectQueryAnalyzer::InterpreterSelectQueryAnalyzer(std::shared_ptr const&, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&) @ 0x000000001785f7e0 E 8. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:701: std::unique_ptr> std::__function::__policy_invoker> (DB::InterpreterFactory::Arguments const&)>::__call_impl> (DB::InterpreterFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::InterpreterFactory::Arguments const&) @ 0x0000000017864110 E 9. ./build_docker/./src/Interpreters/InterpreterFactory.cpp:0: DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x00000000177cfb6b E 10. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:302: DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x0000000017dbf13f E 11. ./build_docker/./src/Interpreters/executeQuery.cpp:1376: DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000017dbac38 E 12. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:612: DB::TCPHandler::runImpl() @ 0x00000000195edc92 E 13. ./build_docker/./src/Server/TCPHandler.cpp:2341: DB::TCPHandler::run() @ 0x000000001960c2e8 E 14. ./build_docker/./base/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x000000001d16e0a3 E 15. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x000000001d16e912 E 16. ./build_docker/./base/poco/Foundation/src/ThreadPool.cpp:202: Poco::PooledThread::run() @ 0x000000001d3705c7 E 17. ./build_docker/./base/poco/Foundation/src/Thread.cpp:46: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x000000001d36e890 E 18. ./base/poco/Foundation/include/Poco/SharedPtr.h:231: Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001d36cd4a E 19. __tsan_thread_start_func @ 0x000000000706b0cf E 20. ? @ 0x00007fbf65c25ac3 E 21. ? @ 0x00007fbf65cb7850 E . (UNKNOWN_TABLE) E (query: select * from `test_database.postgresql_replica_0` order by key;) helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- PostgreSQL is available - running test ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_0" ( key Integer NOT NULL, value1 Integer, value2 Integer, value3 Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_1" ( key Integer NOT NULL, value1 Integer, value2 Integer, value3 Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_2" ( key Integer NOT NULL, value1 Integer, value2 Integer, value3 Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_3" ( key Integer NOT NULL, value1 Integer, value2 Integer, value3 Integer, PRIMARY KEY(key)) Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_4" ( key Integer NOT NULL, value1 Integer, value2 Integer, value3 Integer, PRIMARY KEY(key)) Checking table postgresql_replica_0 exists in test_database Checking table is synchronized: test_database.postgresql_replica_0 ------------------------------ Captured log call ------------------------------- 2025-06-08 17:57:14 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_0 SELECT number, 0, 0, 0 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:15 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_1 SELECT number, 1, 1, 1 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:15 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_2 SELECT number, 2, 2, 2 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:15 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_3 SELECT number, 3, 3, 3 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:15 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_4 SELECT number, 4, 4, 4 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:16 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:16 [ 550 ] DEBUG : Executing query CREATE DATABASE `test_database` ENGINE = MaterializedPostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) 2025-06-08 17:57:16 [ 550 ] DEBUG : Executing query SHOW DATABASES on instance (cluster.py:3455, query) 2025-06-08 17:57:16 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_0 SELECT 25 + number, 0, 0, 0 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:16 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_1 SELECT 25 + number, 1, 1, 1 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:16 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_2 SELECT 25 + number, 2, 2, 2 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:17 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_3 SELECT 25 + number, 3, 3, 3 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:17 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_4 SELECT 25 + number, 4, 4, 4 from numbers(25) on instance (cluster.py:3455, query) 2025-06-08 17:57:17 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:17 [ 550 ] DEBUG : Executing query select * from `postgres_database`.`postgresql_replica_0` order by key; on instance (cluster.py:3455, query) 2025-06-08 17:57:18 [ 550 ] DEBUG : Executing query select * from `test_database.postgresql_replica_0` order by key; on instance (cluster.py:3455, query) ---------------------------- Captured log teardown ----------------------------- 2025-06-08 17:57:18 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` SYNC on instance (cluster.py:3455, query) 2025-06-08 17:57:18 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:18 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:18 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) ___________________________ test_user_managed_slots ____________________________ started_cluster = def test_user_managed_slots(started_cluster): slot_name = "user_slot" table_name = "test_table" pg_manager.create_and_fill_postgres_table(table_name) replication_connection = get_postgres_conn( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, database=True, replication=True, auto_commit=True, ) snapshot = create_replication_slot(replication_connection, slot_name=slot_name) pg_manager.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, settings=[ f"materialized_postgresql_replication_slot = '{slot_name}'", f"materialized_postgresql_snapshot = '{snapshot}'", ], ) > check_tables_are_synchronized(instance, table_name) test_postgresql_replica_database_engine_1/test.py:865: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:392: in check_tables_are_synchronized result = instance.query(result_query) helpers/cluster.py:3456: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 60, stderr: Received exception from server (version 24.3.18): E Code: 60. DB::Exception: Received from 172.16.1.3:9000. DB::Exception: Unknown table expression identifier 'test_database.test_table' in scope SELECT * FROM `test_database.test_table` ORDER BY key ASC. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000001d2e15a3 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000eff24b4 E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x0000000007d363fc E 3. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::(anonymous namespace)::QueryAnalyzer::resolveQuery(std::shared_ptr const&, DB::(anonymous namespace)::IdentifierResolveScope&) @ 0x00000000173a87a6 E 4. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::QueryAnalysisPass::run(std::shared_ptr&, std::shared_ptr) @ 0x00000000173a162f E 5. ./build_docker/./src/Analyzer/QueryTreePassManager.cpp:0: DB::QueryTreePassManager::run(std::shared_ptr) @ 0x000000001739ca9a E 6. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:0: DB::(anonymous namespace)::buildQueryTreeAndRunPasses(std::shared_ptr const&, DB::SelectQueryOptions const&, std::shared_ptr const&, std::shared_ptr const&) @ 0x0000000017862e10 E 7. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:160: DB::InterpreterSelectQueryAnalyzer::InterpreterSelectQueryAnalyzer(std::shared_ptr const&, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&) @ 0x000000001785f7e0 E 8. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:701: std::unique_ptr> std::__function::__policy_invoker> (DB::InterpreterFactory::Arguments const&)>::__call_impl> (DB::InterpreterFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::InterpreterFactory::Arguments const&) @ 0x0000000017864110 E 9. ./build_docker/./src/Interpreters/InterpreterFactory.cpp:0: DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x00000000177cfb6b E 10. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:302: DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x0000000017dbf13f E 11. ./build_docker/./src/Interpreters/executeQuery.cpp:1376: DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000017dbac38 E 12. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:612: DB::TCPHandler::runImpl() @ 0x00000000195edc92 E 13. ./build_docker/./src/Server/TCPHandler.cpp:2341: DB::TCPHandler::run() @ 0x000000001960c2e8 E 14. ./build_docker/./base/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x000000001d16e0a3 E 15. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x000000001d16e912 E 16. ./build_docker/./base/poco/Foundation/src/ThreadPool.cpp:202: Poco::PooledThread::run() @ 0x000000001d3705c7 E 17. ./build_docker/./base/poco/Foundation/src/Thread.cpp:46: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x000000001d36e890 E 18. ./base/poco/Foundation/include/Poco/SharedPtr.h:231: Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001d36cd4a E 19. __tsan_thread_start_func @ 0x000000000706b0cf E 20. ? @ 0x00007fbf65c25ac3 E 21. ? @ 0x00007fbf65cb7850 E . (UNKNOWN_TABLE) E (query: select * from `test_database.test_table` order by key;) helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- PostgreSQL is available - running test ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "test_table" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) user_slot 0/82C2058 000000DA-00000002-1 Checking table test_table exists in test_database Checking table is synchronized: test_database.test_table ------------------------------ Captured log call ------------------------------- 2025-06-08 17:57:19 [ 550 ] DEBUG : Executing query INSERT INTO `postgres_database`.`test_table` SELECT number, number from numbers(50) on instance (cluster.py:3455, query) 2025-06-08 17:57:19 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:19 [ 550 ] DEBUG : Executing query CREATE DATABASE `test_database` ENGINE = MaterializedPostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') SETTINGS materialized_postgresql_replication_slot = 'user_slot', materialized_postgresql_snapshot = '000000DA-00000002-1' on instance (cluster.py:3455, query) 2025-06-08 17:57:19 [ 550 ] DEBUG : Executing query SHOW DATABASES on instance (cluster.py:3455, query) 2025-06-08 17:57:19 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:20 [ 550 ] DEBUG : Executing query select * from `postgres_database`.`test_table` order by key; on instance (cluster.py:3455, query) 2025-06-08 17:57:20 [ 550 ] DEBUG : Executing query select * from `test_database.test_table` order by key; on instance (cluster.py:3455, query) ---------------------------- Captured log teardown ----------------------------- 2025-06-08 17:57:20 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` SYNC on instance (cluster.py:3455, query) 2025-06-08 17:57:20 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:21 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:21 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) _____________________________ test_virtual_columns _____________________________ started_cluster = def test_virtual_columns(started_cluster): conn = get_postgres_conn( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, database=True, ) cursor = conn.cursor() table_name = "postgresql_replica_0" create_postgres_table(cursor, table_name) pg_manager.create_materialized_db( ip=started_cluster.postgres_ip, port=started_cluster.postgres_port, ) assert_nested_table_is_created(instance, table_name) instance.query( f"INSERT INTO postgres_database.{table_name} SELECT number, number from numbers(10)" ) > check_tables_are_synchronized(instance, table_name) test_postgresql_replica_database_engine_1/test.py:553: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ helpers/postgres_utility.py:392: in check_tables_are_synchronized result = instance.query(result_query) helpers/cluster.py:3456: in query return self.client.query( helpers/client.py:36: in wrap return func(self, *args, **kwargs) helpers/client.py:74: in query ).get_answer() _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = def get_answer(self): self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) self.stdout_file.seek(0) self.stderr_file.seek(0) stdout = self.stdout_file.read().decode("utf-8", errors="replace") stderr = self.stderr_file.read().decode("utf-8", errors="replace") if ( self.timer is not None and not self.process_finished_before_timeout and not self.ignore_error ): logging.debug(f"Timed out. Last stdout:{stdout}, stderr:{stderr}") raise QueryTimeoutExceedException("Client timed out!") if ( self.process.returncode != 0 or self.remove_trash_from_stderr(stderr) ) and not self.ignore_error: > raise QueryRuntimeException( "Client failed! Return code: {}, stderr: {}".format( self.process.returncode, stderr ), self.process.returncode, stderr, ) E helpers.client.QueryRuntimeException: Client failed! Return code: 60, stderr: Received exception from server (version 24.3.18): E Code: 60. DB::Exception: Received from 172.16.1.3:9000. DB::Exception: Unknown table expression identifier 'test_database.postgresql_replica_0' in scope SELECT * FROM `test_database.postgresql_replica_0` ORDER BY key ASC. Stack trace: E E 0. ./contrib/llvm-project/libcxx/include/exception:141: Poco::Exception::Exception(String const&, int) @ 0x000000001d2e15a3 E 1. ./build_docker/./src/Common/Exception.cpp:96: DB::Exception::Exception(DB::Exception::MessageMasked&&, int, bool) @ 0x000000000eff24b4 E 2. DB::Exception::Exception(int, FormatStringHelperImpl::type, std::type_identity::type>, String const&, String&&) @ 0x0000000007d363fc E 3. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::(anonymous namespace)::QueryAnalyzer::resolveQuery(std::shared_ptr const&, DB::(anonymous namespace)::IdentifierResolveScope&) @ 0x00000000173a87a6 E 4. ./build_docker/./src/Analyzer/Passes/QueryAnalysisPass.cpp:0: DB::QueryAnalysisPass::run(std::shared_ptr&, std::shared_ptr) @ 0x00000000173a162f E 5. ./build_docker/./src/Analyzer/QueryTreePassManager.cpp:0: DB::QueryTreePassManager::run(std::shared_ptr) @ 0x000000001739ca9a E 6. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:0: DB::(anonymous namespace)::buildQueryTreeAndRunPasses(std::shared_ptr const&, DB::SelectQueryOptions const&, std::shared_ptr const&, std::shared_ptr const&) @ 0x0000000017862e10 E 7. ./build_docker/./src/Interpreters/InterpreterSelectQueryAnalyzer.cpp:160: DB::InterpreterSelectQueryAnalyzer::InterpreterSelectQueryAnalyzer(std::shared_ptr const&, std::shared_ptr const&, DB::SelectQueryOptions const&, std::vector> const&) @ 0x000000001785f7e0 E 8. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:701: std::unique_ptr> std::__function::__policy_invoker> (DB::InterpreterFactory::Arguments const&)>::__call_impl> (DB::InterpreterFactory::Arguments const&)>>(std::__function::__policy_storage const*, DB::InterpreterFactory::Arguments const&) @ 0x0000000017864110 E 9. ./build_docker/./src/Interpreters/InterpreterFactory.cpp:0: DB::InterpreterFactory::get(std::shared_ptr&, std::shared_ptr, DB::SelectQueryOptions const&) @ 0x00000000177cfb6b E 10. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:302: DB::executeQueryImpl(char const*, char const*, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum, DB::ReadBuffer*) @ 0x0000000017dbf13f E 11. ./build_docker/./src/Interpreters/executeQuery.cpp:1376: DB::executeQuery(String const&, std::shared_ptr, DB::QueryFlags, DB::QueryProcessingStage::Enum) @ 0x0000000017dbac38 E 12. ./contrib/llvm-project/libcxx/include/__memory/shared_ptr.h:612: DB::TCPHandler::runImpl() @ 0x00000000195edc92 E 13. ./build_docker/./src/Server/TCPHandler.cpp:2341: DB::TCPHandler::run() @ 0x000000001960c2e8 E 14. ./build_docker/./base/poco/Net/src/TCPServerConnection.cpp:57: Poco::Net::TCPServerConnection::start() @ 0x000000001d16e0a3 E 15. ./contrib/llvm-project/libcxx/include/__memory/unique_ptr.h:48: Poco::Net::TCPServerDispatcher::run() @ 0x000000001d16e912 E 16. ./build_docker/./base/poco/Foundation/src/ThreadPool.cpp:202: Poco::PooledThread::run() @ 0x000000001d3705c7 E 17. ./build_docker/./base/poco/Foundation/src/Thread.cpp:46: Poco::(anonymous namespace)::RunnableHolder::run() @ 0x000000001d36e890 E 18. ./base/poco/Foundation/include/Poco/SharedPtr.h:231: Poco::ThreadImpl::runnableEntry(void*) @ 0x000000001d36cd4a E 19. __tsan_thread_start_func @ 0x000000000706b0cf E 20. ? @ 0x00007fbf65c25ac3 E 21. ? @ 0x00007fbf65cb7850 E . (UNKNOWN_TABLE) E (query: select * from `test_database.postgresql_replica_0` order by key;) helpers/client.py:239: QueryRuntimeException ---------------------------- Captured stdout setup ----------------------------- PostgreSQL is available - running test ----------------------------- Captured stdout call ----------------------------- Query: CREATE TABLE IF NOT EXISTS "postgresql_replica_0" ( key Integer NOT NULL, value Integer, PRIMARY KEY(key)) Checking table postgresql_replica_0 exists in test_database Checking table postgresql_replica_0 exists in test_database Checking table is synchronized: test_database.postgresql_replica_0 ------------------------------ Captured log call ------------------------------- 2025-06-08 17:57:21 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:21 [ 550 ] DEBUG : Executing query CREATE DATABASE `test_database` ENGINE = MaterializedPostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) 2025-06-08 17:57:21 [ 550 ] DEBUG : Executing query SHOW DATABASES on instance (cluster.py:3455, query) 2025-06-08 17:57:21 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:22 [ 550 ] DEBUG : Executing query INSERT INTO postgres_database.postgresql_replica_0 SELECT number, number from numbers(10) on instance (cluster.py:3455, query) 2025-06-08 17:57:22 [ 550 ] DEBUG : Executing query SHOW TABLES FROM `test_database` on instance (cluster.py:3455, query) 2025-06-08 17:57:22 [ 550 ] DEBUG : Executing query select * from `postgres_database`.`postgresql_replica_0` order by key; on instance (cluster.py:3455, query) 2025-06-08 17:57:22 [ 550 ] DEBUG : Executing query select * from `test_database.postgresql_replica_0` order by key; on instance (cluster.py:3455, query) ---------------------------- Captured log teardown ----------------------------- 2025-06-08 17:57:23 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS `test_database` SYNC on instance (cluster.py:3455, query) 2025-06-08 17:57:23 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:23 [ 550 ] DEBUG : Executing query DROP DATABASE IF EXISTS "postgres_database" on instance (cluster.py:3455, query) 2025-06-08 17:57:23 [ 550 ] DEBUG : Executing query CREATE DATABASE "postgres_database" ENGINE = PostgreSQL('172.16.1.2:5432', 'postgres_database', 'postgres', 'mysecretpassword') on instance (cluster.py:3455, query) 2025-06-08 17:57:24 [ 550 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine1', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', 'stop', '--timeout', '20'] (cluster.py:113, run_and_check) 2025-06-08 17:57:24 [ 550 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine1_instance_1 ... (cluster.py:123, run_and_check) 2025-06-08 17:57:24 [ 550 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine1_postgres1_1 ... (cluster.py:123, run_and_check) 2025-06-08 17:57:24 [ 550 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine1_postgres1_1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:57:24 [ 550 ] DEBUG : Stderr:Stopping roottestpostgresqlreplicadatabaseengine1_instance_1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:57:24 [ 550 ] DEBUG : Command:['bash', '-c', '[ -f /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/logs/stderr.log* || true'] (cluster.py:113, run_and_check) 2025-06-08 17:57:24 [ 550 ] DEBUG : Command:['docker-compose', '--env-file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/.env', '--project-name', 'roottestpostgresqlreplicadatabaseengine1', '--file', '/ClickHouse/tests/integration/test_postgresql_replica_database_engine_1/_instances_1/instance/docker-compose.yml', '--file', '/ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_postgres.yml', 'down', '--volumes'] (cluster.py:113, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine1_instance_1 ... (cluster.py:123, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine1_postgres1_1 ... (cluster.py:123, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine1_instance_1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stderr:Removing roottestpostgresqlreplicadatabaseengine1_postgres1_1 ... done (cluster.py:123, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stderr:Removing network roottestpostgresqlreplicadatabaseengine1_default (cluster.py:123, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Cleanup called (cluster.py:801, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : Docker networks for project roottestpostgresqlreplicadatabaseengine1 are NETWORK ID NAME DRIVER SCOPE (cluster.py:780, print_all_docker_pieces) 2025-06-08 17:57:25 [ 550 ] DEBUG : Docker containers for project roottestpostgresqlreplicadatabaseengine1 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:788, print_all_docker_pieces) 2025-06-08 17:57:25 [ 550 ] DEBUG : Docker volumes for project roottestpostgresqlreplicadatabaseengine1 are DRIVER VOLUME NAME (cluster.py:796, print_all_docker_pieces) 2025-06-08 17:57:25 [ 550 ] DEBUG : Command:docker container list --all --filter name='^/roottestpostgresqlreplicadatabaseengine1_.*_1$' --format '{{.ID}}:{{.Names}}' (cluster.py:113, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Unstopped containers: {} (cluster.py:815, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : No running containers for project: roottestpostgresqlreplicadatabaseengine1 (cluster.py:829, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : Trying to prune unused networks... (cluster.py:835, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : Trying to prune unused images... (cluster.py:851, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : Command:['docker', 'image', 'prune', '-f'] (cluster.py:113, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:121, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Images pruned (cluster.py:854, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : Trying to prune unused volumes... (cluster.py:860, cleanup) 2025-06-08 17:57:25 [ 550 ] DEBUG : Command:['docker volume ls | wc -l'] (cluster.py:113, run_and_check) 2025-06-08 17:57:25 [ 550 ] DEBUG : Stdout:1 (cluster.py:121, run_and_check) ============================== slowest durations =============================== 15.98s setup test_postgresql_replica_database_engine_1/test.py::test_replicating_dml 13.42s call test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished 7.51s call test_postgresql_replica_database_engine_1/test.py::test_single_transaction 3.58s call test_postgresql_replica_database_engine_1/test.py::test_replicating_dml 3.28s call test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes 2.54s teardown test_postgresql_replica_database_engine_1/test.py::test_virtual_columns 1.58s call test_postgresql_replica_database_engine_1/test.py::test_virtual_columns 1.38s call test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots 0.84s teardown test_postgresql_replica_database_engine_1/test.py::test_single_transaction 0.83s teardown test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots 0.74s teardown test_postgresql_replica_database_engine_1/test.py::test_replicating_dml 0.73s teardown test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished 0.73s teardown test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes 0.00s setup test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots 0.00s setup test_postgresql_replica_database_engine_1/test.py::test_single_transaction 0.00s setup test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes 0.00s setup test_postgresql_replica_database_engine_1/test.py::test_virtual_columns 0.00s setup test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished =========================== short test summary info ============================ FAILED test_postgresql_replica_database_engine_1/test.py::test_replicating_dml FAILED test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished FAILED test_postgresql_replica_database_engine_1/test.py::test_single_transaction FAILED test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes FAILED test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots FAILED test_postgresql_replica_database_engine_1/test.py::test_virtual_columns ============================== 6 failed in 53.85s ============================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 437, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_ikbgrs --privileged --dns-search='.' --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=2cffe1eae894 -e DOCKER_BASE_TAG=2993bc2bf171 -e DOCKER_KERBERIZED_HADOOP_TAG=ce74919e88f5 -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=a2d3dc777d0c -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS=" -rfEps --run-id=1 --color=no --durations=0 test_postgresql_replica_database_engine_1/test.py::test_replicating_dml test_postgresql_replica_database_engine_1/test.py::test_restart_server_while_replication_startup_not_finished test_postgresql_replica_database_engine_1/test.py::test_single_transaction test_postgresql_replica_database_engine_1/test.py::test_table_schema_changes test_postgresql_replica_database_engine_1/test.py::test_user_managed_slots test_postgresql_replica_database_engine_1/test.py::test_virtual_columns -vvv" altinityinfra/integration-tests-runner:9d492c2eec24 ' returned non-zero exit status 1.